v2.29.0 #5359
mudler
announced in
Announcements
v2.29.0
#5359
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
v2.29.0
I am thrilled to announce the release of LocalAI v2.29.0! This update focuses heavily on refining our container image strategy, making default images leaner and providing clearer options for users needing specific features or hardware acceleration. We've also added support for new models like Qwen3, enhanced existing backends, and introduced experimental endpoints, like video generation!
This release includes significant changes to container image tagging and contents. Please review carefully:
diffusers
) now require the-extras
suffix (e.g.,latest-gpu-nvidia-cuda-12-extras
). Default images are now slimmer and do not include these dependencies.-ffmpeg
tags have been removed. If you previously used an-ffmpeg
tagged image, simply switch to the corresponding base image tag (e.g.,latest-gpu-hipblas-ffmpeg
becomeslatest-gpu-hipblas
).Here below some examples, note that the CI is still publishing the images so won't be available until jobs are processed, and the installation scripts will be updated right after images are publicly available.
CPU only image:
NVIDIA GPU Images:
AMD GPU Images (ROCm):
Intel GPU Images (oneAPI):
Vulkan GPU Images:
# Vulkan with core features docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-gpu-vulkan
AIO Images (pre-downloaded models):
For more information about the AIO images and pre-downloaded models, see Container Documentation.
Key Changes in v2.29.0
📦 Container Image Overhaul
-extras
Suffix: Images with additional Python dependencies are now identified by the-extras
suffix.latest
,latest-gpu-nvidia-cuda-12
) now provide core LocalAI functionality without the extra Python libraries.latest-*
Tags: Added specificlatest
tags for various GPU architectures:latest-gpu-hipblas
(AMD ROCm)latest-gpu-intel-f16
(Intel oneAPI FP16)latest-gpu-intel-f32
(Intel oneAPI FP32)latest-gpu-nvidia-cuda-12
(NVIDIA CUDA 12)latest-gpu-vulkan
(Vulkan)🚀 New Features & Enhancements
llama.cpp
andCLIP
./video/generations
endpoint. Stay tuned for compatible model backends!install.sh
script now includes a--uninstall
flag for easy removal.gfx803,gfx900,gfx906,gfx908,gfx90a,gfx942,gfx1010,gfx1030,gfx1032,gfx1100,gfx1101,gfx1102
🧹 Backend Updates
The Complete Local Stack for Privacy-First AI
With LocalAGI rejoining LocalAI alongside LocalRecall, our ecosystem provides a complete, open-source stack for private, secure, and intelligent AI operations:
LocalAI
The free, Open Source OpenAI alternative. Acts as a drop-in replacement REST API compatible with OpenAI specifications for local AI inferencing. No GPU required.
Link: https://github.com/mudler/LocalAI
LocalAGI
A powerful Local AI agent management platform. Serves as a drop-in replacement for OpenAI's Responses API, supercharged with advanced agentic capabilities and a no-code UI.
Link: https://github.com/mudler/LocalAGI
LocalRecall
A RESTful API and knowledge base management system providing persistent memory and storage capabilities for AI agents. Designed to work alongside LocalAI and LocalAGI.
Link: https://github.com/mudler/LocalRecall
Join the Movement! ❤️
A massive THANK YOU to our incredible community! LocalAI has over 31,800 stars, and LocalAGI has already rocketed past 450+ stars!
As a reminder, LocalAI is real FOSS (Free and Open Source Software) and its sibling projects are community-driven and not backed by VCs or a company. We rely on contributors donating their spare time. If you love open-source, privacy-first AI, please consider starring the repos, contributing code, reporting bugs, or spreading the word!
👉 Check out the reborn LocalAGI v2 today: https://github.com/mudler/LocalAGI
Let's continue building the future of AI, together! 🙌
Full changelog 👇
👉 Click to expand 👈
What's Changed
Breaking Changes 🛠
Bug fixes 🐛
Exciting New Features 🎉
🧠 Models
📖 Documentation and examples
👒 Dependencies
80f19b41869728eeb6a26569957b92a773a2b2c6
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to80f19b41869728eeb6a26569957b92a773a2b2c6
#5183015022bb53387baa8b23817ac03743705c7d472b
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to015022bb53387baa8b23817ac03743705c7d472b
#51922f74c354c0f752ed9aabf7d3a350e6edebd7e744
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to2f74c354c0f752ed9aabf7d3a350e6edebd7e744
#52036408210082cc0a61b992b487be7e2ff2efbb9e36
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to6408210082cc0a61b992b487be7e2ff2efbb9e36
#521100137157fca3d17b90380762b4d7cc158d385bd3
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to00137157fca3d17b90380762b4d7cc158d385bd3
#52186602304814e679cc8c162bb760a034aceb4f8965
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to6602304814e679cc8c162bb760a034aceb4f8965
#52281d735c0b4fa0551c51c2f4ac888dd9a01f447985
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to1d735c0b4fa0551c51c2f4ac888dd9a01f447985
#5233658987cfc9d752dca7758987390d5fb1a7a0a54a
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to658987cfc9d752dca7758987390d5fb1a7a0a54a
#5234ecda2ec4b347031a9b8a89ee2efc664ce63f599c
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toecda2ec4b347031a9b8a89ee2efc664ce63f599c
#5238226251ed56b85190e18a1cca963c45b888f4953c
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to226251ed56b85190e18a1cca963c45b888f4953c
#5240295354ea6848a77bdee204ee1c971d9b92ffcca9
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to295354ea6848a77bdee204ee1c971d9b92ffcca9
#524577d5e9a76a7b4a8a7c5bf9cf6ebef91860123cba
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to77d5e9a76a7b4a8a7c5bf9cf6ebef91860123cba
#5254ced44be34290fab450f8344efa047d8a08e723b4
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toced44be34290fab450f8344efa047d8a08e723b4
#52585f5e39e1ba5dbea814e41f2a15e035d749a520bc
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to5f5e39e1ba5dbea814e41f2a15e035d749a520bc
#5267e2e1ddb93a01ce282e304431b37e60b3cddb6114
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toe2e1ddb93a01ce282e304431b37e60b3cddb6114
#52783e168bede4d27b35656ab8026015b87659ecbec2
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to3e168bede4d27b35656ab8026015b87659ecbec2
#5284d7a14c42a1883a34a6553cbfe30da1e1b84dfd6a
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp tod7a14c42a1883a34a6553cbfe30da1e1b84dfd6a
#529236667c8edcded08063ed51c7d57e9e086bbfc903
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to36667c8edcded08063ed51c7d57e9e086bbfc903
#53009fdfcdaeddd1ef57c6d041b89cd8fb7048a0f028
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to9fdfcdaeddd1ef57c6d041b89cd8fb7048a0f028
#5316b34c859146630dff136943abc9852ca173a7c9d6
by @mudler in chore(deps): bump llama.cpp tob34c859146630dff136943abc9852ca173a7c9d6
#532391a86a6f354aa73a7aab7bc3d283be410fdc93a5
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to91a86a6f354aa73a7aab7bc3d283be410fdc93a5
#5329814f795e063c257f33b921eab4073484238a151a
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to814f795e063c257f33b921eab4073484238a151a
#5331f05a6d71a0f3dbf0730b56a1abbad41c0f42e63d
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp tof05a6d71a0f3dbf0730b56a1abbad41c0f42e63d
#534033eff4024084d1f0c8441b79f7208a52fad79858
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to33eff4024084d1f0c8441b79f7208a52fad79858
#534315e6125a397f6086c1dfdf7584acdb7c730313dc
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to15e6125a397f6086c1dfdf7584acdb7c730313dc
#53452e310b841e0b4e7cf00890b53411dd9f8578f243
by @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp to2e310b841e0b4e7cf00890b53411dd9f8578f243
#47859a390c4829cd3058d26a2e2c09d16e3fd12bf1b1
by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to9a390c4829cd3058d26a2e2c09d16e3fd12bf1b1
#5351Other Changes
New Contributors
Full Changelog: v2.28.0...v2.29.0
This discussion was created from the release v2.29.0.
Beta Was this translation helpful? Give feedback.
All reactions